Skip to content

Conversation

@tjhunter
Copy link
Collaborator

Description

./packages/evaluate/src/weathergen/evaluate/run_evaluation.py

Issue Number

Refs #1092

Is this PR a draft? Mark it as draft.

Checklist before asking for review

  • I have performed a self-review of my code
  • My changes comply with basic sanity checks:
    • I have fixed formatting issues with ./scripts/actions.sh lint
    • I have run unit tests with ./scripts/actions.sh unit-test
    • I have documented my code and I have updated the docstrings.
    • I have added unit tests, if relevant
  • I have tried my changes with data and code:
    • I have run the integration tests with ./scripts/actions.sh integration-test
    • (bigger changes) I have run a full training and I have written in the comment the run_id(s): launch-slurm.py --time 60
    • (bigger changes and experiments) I have shared a hegdedoc in the github issue with all the configurations and runs for this experiments
  • I have informed and aligned with people impacted by my change:
    • for config changes: the MatterMost channels and/or a design doc
    • for changes of dependencies: the MatterMost software development channel

readme = "../../README.md"
requires-python = ">=3.12,<3.13"
dependencies = [
"mlflow",
Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

maybe mlflow-skinny would work

@grassesi grassesi closed this Oct 27, 2025
@grassesi grassesi deleted the tjh/dev/1092_mlflow branch October 27, 2025 13:47
@grassesi grassesi restored the tjh/dev/1092_mlflow branch October 27, 2025 16:37
@tjhunter tjhunter reopened this Oct 29, 2025
channels_set = collect_channels(scores_dict, metric, region, runs)

for run_id, metrics_dict in reordered_dict.items():
parent_run = get_or_create_mlflow_parent_run(mlflow_client, run_id)
Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

the problem is the run_id here, which is the inference run.the easiest is probably to query mlflow to find the value of the tag tags.from_run_id (which will be the model).

This or reading the config for this run. I think it will be more complicated.

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I pushed an update, getting from_run_id from the run config and using it to generate parent_run.
Upon uploading the scores, ordering on MLFlow looks as below:
image
We have to check if everything else is in order in terms of relations between run_ids.

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Not sure if we should keep the inference run_id in the name or rather have it in the metadata. On the other hand might be good to keep it in the name in case of several inference instances...

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

Status: Done

Development

Successfully merging this pull request may close these issues.

4 participants